首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   100038篇
  免费   4131篇
  国内免费   2963篇
化学   39949篇
晶体学   1044篇
力学   8007篇
综合类   242篇
数学   34331篇
物理学   23559篇
  2023年   492篇
  2022年   511篇
  2021年   720篇
  2020年   976篇
  2019年   890篇
  2018年   10977篇
  2017年   10780篇
  2016年   7012篇
  2015年   1807篇
  2014年   1517篇
  2013年   1917篇
  2012年   5776篇
  2011年   12417篇
  2010年   6962篇
  2009年   7304篇
  2008年   8014篇
  2007年   10034篇
  2006年   1538篇
  2005年   2355篇
  2004年   2364篇
  2003年   2661篇
  2002年   1609篇
  2001年   754篇
  2000年   797篇
  1999年   744篇
  1998年   679篇
  1997年   564篇
  1996年   609篇
  1995年   499篇
  1994年   429篇
  1993年   413篇
  1992年   318篇
  1991年   291篇
  1990年   281篇
  1989年   215篇
  1988年   177篇
  1987年   139篇
  1986年   153篇
  1985年   112篇
  1984年   79篇
  1983年   75篇
  1982年   78篇
  1981年   58篇
  1980年   54篇
  1979年   48篇
  1914年   45篇
  1913年   40篇
  1912年   40篇
  1909年   41篇
  1908年   40篇
排序方式: 共有10000条查询结果,搜索用时 312 毫秒
111.
Frank Jackson and Philip Pettit have defended a non-reductive account of causal relevance known as the ‘program explanation account’. Allegedly, irreducible mental properties can be causally relevant in virtue of figuring in non-redundant program explanations which convey information not conveyed by explanations in terms of the physical properties that actually do the ‘causal work’. I argue that none of the possible ways to spell out the intuitively plausible idea of a program explanation serves its purpose, viz., defends non-reductive physicalism against Jaegwon Kim’s Causal Exclusion Argument according to which non-reductive physicalism is committed to epiphenomenalism because irreducible mental properties are ‘screened off’ from causal relevance by their physical realizers. Jackson and Pettit’s most promising explication of a program explanation appeals to the idea of invariance of effect under variation of realization, but I show that invariance of effect under variation of realization is neither necessary nor sufficient for causal relevance.  相似文献   
112.
113.
For 1 ≤ dk, let Kk/d be the graph with vertices 0, 1, …, k ? 1, in which ij if d ≤ |i ? j| ≤ k ? d. The circular chromatic number χc(G) of a graph G is the minimum of those k/d for which G admits a homomorphism to Kk/d. The circular clique number ωc(G) of G is the maximum of those k/d for which Kk/d admits a homomorphism to G. A graph G is circular perfect if for every induced subgraph H of G, we have χc(H) = ωc(H). In this paper, we prove that if G is circular perfect then for every vertex x of G, NG[x] is a perfect graph. Conversely, we prove that if for every vertex x of G, NG[x] is a perfect graph and G ? N[x] is a bipartite graph with no induced P5 (the path with five vertices), then G is a circular perfect graph. In a companion paper, we apply the main result of this paper to prove an analog of Haj?os theorem for circular chromatic number for k/d ≥ 3. Namely, we shall design a few graph operations and prove that for any k/d ≥ 3, starting from the graph Kk/d, one can construct all graphs of circular chromatic number at least k/d by repeatedly applying these graph operations. © 2005 Wiley Periodicals, Inc. J Graph Theory 48: 186–209, 2005  相似文献   
114.
We develop a method of randomizing units to treatments that relies on subjective judgement or on possible coarse modeling to produce restrictions on the randomization. The procedure thus fits within the general framework of ranked set sampling. However, instead of selecting a single unit from each set for full measurement, all units within a set are used. The units within a set are assigned to different treatments. Such an assignment translates the positive dependence among units within a set into a reduction in variation of contrasting features of the treatments. A test for treatment versus control comparison, with controlled familywise error rate, is developed along with the associated confidence intervals. The new procedure is shown to be superior to corresponding procedures based on completely randomized or ranked set sample designs. The superiority appears both in asymptotic relative efficiency and in power for finite sample sizes. Importantly, this test does not rely on perfect rankings; rather, the information in the data on the quality of rankings is exploited to maintain the level of the test when rankings are imperfect. The asymptotic relative efficiency of the test is not affected by estimation of the quality of rankings, and the finite sample performance is only mildly affected.  相似文献   
115.
116.
What strategy should a football (soccer, in American parlance) club adopt when deciding whether to sack its manager? This paper introduces a simple model assuming that a club's objective is to maximize the number of league points that it scores per season. The club's strategy consists of three choices: the length of the honeymoon period during which it will not consider sacking a new manager, the level of the performance trapdoor below which the manager get the sack, and the weight that it will give to more recent games compared to earlier ones. Some data from the last six seasons of the English Premiership are used to calibrate the model. At this early stage of the research, the best strategy appears to have only a short honeymoon period of eight games (much less than the actual shortest period of 12 games), to set the trapdoor at 0.74 points per game, and to put 47% of the weight on the last five games. A club adopting this strategy would obtain on average 56.8 points per season, compared to a Premiership average of 51.8 points.  相似文献   
117.
118.
We consider estimation of loss for generalized Bayes or pseudo-Bayes estimators of a multivariate normal mean vector, θ. In 3 and higher dimensions, the MLEX is UMVUE and minimax but is inadmissible. It is dominated by the James-Stein estimator and by many others. Johnstone (1988, On inadmissibility of some unbiased estimates of loss,Statistical Decision Theory and Related Topics, IV (eds. S. S. Gupta and J. O. Berger), Vol. 1, 361–379, Springer, New York) considered the estimation of loss for the usual estimatorX and the James-Stein estimator. He found improvements over the Stein unbiased estimator of risk. In this paper, for a generalized Bayes point estimator of θ, we compare generalized Bayes estimators to unbiased estimators of loss. We find, somewhat surprisingly, that the unbiased estimator often dominates the corresponding generalized Bayes estimator of loss for priors which give minimax estimators in the original point estimation problem. In particular, we give a class of priors for which the generalized Bayes estimator of θ is admissible and minimax but for which the unbiased estimator of loss dominates the generalized Bayes estimator of loss. We also give a general inadmissibility result for a generalized Bayes estimator of loss. Research supported by NSF Grant DMS-97-04524.  相似文献   
119.
在程序升温条件下 ,用DSC研究了标题化合物的放热分解反应动力学 .用线性最小二乘法、迭代法以及二分法与最小二乘法相结合的方法 ,以积分方程、微分方程和放热速率方程拟合DSC数据 .在逻辑选择建立了微分和积分机理函数的最可几一般表达式后 ,用放热速率方程得到相应的表观活化能 (Ea)、指前因子 (A)和反应级数 (n)的值 .结果表明 :该反应的微分形式的经验动力学模式函数、Ea 和A值分别为 (1-α) 0 .44、2 30 .4kJ/mol和 10 18.16s-1.借助加热速率和所得动力学参数值 ,提出了标题化合物放热分解反应的动力学方程 .该化合物的热爆炸临界温度为 30 2 .6℃ .上述动力学参数对分析、评价标题化合物的稳定性和热变化规律十分有用 .  相似文献   
120.
On effectiveness of wiretap programs in mapping social networks   总被引:1,自引:0,他引:1  
Snowball sampling methods are known to be a biased toward highly connected actors and consequently produce core-periphery networks when these may not necessarily be present. This leads to a biased perception of the underlying network which can have negative policy consequences, as in the identification of terrorist networks. When snowball sampling is used, the potential overload of the information collection system is a distinct problem due to the exponential growth of the number of suspects to be monitored. In this paper, we focus on evaluating the effectiveness of a wiretapping program in terms of its ability to map the rapidly evolving networks within a covert organization. By running a series of simulation-based experiments, we are able to evaluate a broad spectrum of information gathering regimes based on a consistent set of criteria. We conclude by proposing a set of information gathering programs that achieve higher effectiveness then snowball sampling, and at a lower cost. Maksim Tsvetovat is an Assistant Professor at the Center for Social Complexity and department of Public and International Affairs at George Mason University, Fairfax, VA. He received his Ph.D. from the Computation, Organizations and Society program in the School of Computer Science, Carnegie Mellon University. His dissertation was centered on use of artificial intelligence techniques such as planning and semantic reasoning as a means of studying behavior and evolution of complex social networks, such as these of terrorist organizations. He received a Master of Science degree from University of Minnesota with a specialization in Artificial Intelligence and design of Multi-Agent Systems, and has also extensively studied organization theory and social science research methods. His research is centered on building high-fidelity simulations of social and organizational systems using concepts from distributed artificial intelligence and multi-agent systems. Other projects focus on social network analysis for mapping of internal corporate networks or study of covert and terrorist orgnaizations. Maksim’s vita and publications can be found on Kathleen M. Carley is a professor in the School of Computer Science at Carnegie Mellon University and the director of the center for Compuational Analysis of Social and Organizational Systems (CASOS) which has over 25 members, both students and research staff. Her research combines cognitive science, social networks and computer science to address complex social and organizational problems. Her specific research areas are dynamic network analysis, computational social and organization theory, adaptation and evolution, text mining, and the impact of telecommunication technologies and policy on communication, information diffusion, disease contagion and response within and among groups particularly in disaster or crisis situations. She and her lab have developed infrastructure tools for analyzing large scale dynamic networks and various multi-agent simulation systems. The infrastructure tools include ORA, a statistical toolkit for analyzing and visualizing multi-dimensional networks. ORA results are organized into reports that meet various needs such as the management report, the mental model report, and the intelligence report. Another tool is AutoMap, a text-mining systems for extracting semantic networks from texts and then cross-classifying them using an organizational ontology into the underlying social, knowledge, resource and task networks. Her simulation models meld multi-agent technology with network dynamics and empirical data. Three of the large-scale multi-agent network models she and the CASOS group have developed in the counter-terrorism area are: BioWar a city-scale dynamic-network agent-based model for understanding the spread of disease and illness due to natural epidemics, chemical spills, and weaponized biological attacks; DyNet a model of the change in covert networks, naturally and in response to attacks, under varying levels of information uncertainty; and RTE a model for examining state failure and the escalation of conflict at the city, state, nation, and international as changes occur within and among red, blue, and green forces. She is the founding co-editor with Al. Wallace of the journal Computational Organization Theory and has co-edited several books and written over 100 articles in the computational organizations and dynamic network area. Her publications can be found at: http://www.casos.cs.cmu.edu/bios/carley/publications.php  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号